# English Pre-training
T5 Efficient Small Kv32
Apache-2.0
T5-Efficient-SMALL-KV32 is a variant of Google's original T5, adopting a deep narrow architecture focused on improving downstream task performance.
Large Language Model English
T
google
16
0
Testmodel
Apache-2.0
BERT is a transformer model pre-trained on large-scale English corpora through self-supervised learning, utilizing masked language modeling and next sentence prediction objectives
Large Language Model
Transformers English

T
sramasamy8
21
0
T5 Efficient Small Dm768
Apache-2.0
T5-Efficient-SMALL-DM768 is a variant of Google's original T5, adopting a deep narrow architecture that prioritizes increasing model depth to enhance downstream performance.
Large Language Model English
T
google
49
1
T5 Efficient Base Ff9000
Apache-2.0
T5-Efficient-BASE-FF9000 is a variant of Google's original T5, adopting a deep narrow architecture that delivers superior performance on downstream tasks with similar parameter scales.
Large Language Model English
T
google
16
0
T5 Efficient Small Kv256
Apache-2.0
T5-Efficient-SMALL-KV256 is a variant of Google's T5, optimized for downstream task performance using a deep narrow architecture, with 117 million parameters, requiring fine-tuning for use.
Large Language Model English
T
google
16
0
T5 Efficient Mini
Apache-2.0
T5-Efficient-MINI is a variant of Google's original T5, adopting a deep narrow architecture that demonstrates superior downstream task performance among models with similar parameter counts.
Large Language Model English
T
google
946
6
T5 Efficient Tiny Nl2
Apache-2.0
T5-Efficient-TINY-NL2 is a variant of Google's original T5, adopting a deep narrow architecture focused on enhancing downstream task performance.
Large Language Model English
T
google
334
0
T5 Efficient Tiny Nl8
Apache-2.0
T5-Efficient-TINY-NL8 is an efficient variant of the Google T5 model, optimized for downstream task performance using a deep narrow architecture.
Large Language Model English
T
google
25
5
Medium Base
Apache-2.0
A Transformer model pre-trained on English corpus using self-supervised learning similar to ELECTRA, trained by predicting replaced tokens.
Large Language Model
Transformers English

M
funnel-transformer
69
0
T5 Efficient Small El2
Apache-2.0
T5-Efficient-SMALL-EL2 is a variant of Google's original T5, adopting a deep narrow architecture that prioritizes increasing model depth to enhance downstream performance.
Large Language Model English
T
google
20
0
T5 Efficient Large Nh32
Apache-2.0
T5 Efficient Large-NH32 is a deep-narrow variant of Google's T5 model, focusing on improving downstream task performance by increasing model depth.
Large Language Model English
T
google
16
0
T5 Efficient Base Nl48
Apache-2.0
T5-Efficient-BASE-NL48 is a variant of Google T5, adopting a deep narrow architecture that prioritizes increasing model depth to enhance downstream task performance.
Large Language Model English
T
google
14
1
Xlnet Large Cased
MIT
XLNet is an unsupervised language representation learning method based on a generalized permutation language modeling objective, using Transformer-XL as the backbone model, excelling in long-context tasks.
Large Language Model
Transformers English

X
xlnet
2,419
24
Featured Recommended AI Models